Prune Your Model Before Distill It

نویسندگان

چکیده

Knowledge distillation transfers the knowledge from a cumbersome teacher to small student. Recent results suggest that student-friendly is more appropriate distill since it provides transferrable knowledge. In this work, we propose novel framework, “prune, then distill,” prunes model first make and We provide several exploratory examples where pruned teaches better than original unpruned networks. further show theoretically plays role of regularizer in distillation, which reduces generalization error. Based on result, neural network compression scheme student formed based apply distill” strategy. The code available at https://github.com/ososos888/prune-then-distill.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Imagine the Readers of your article before writing it

سخن سردبیر Editorial مجله دانشگاه علوم پزشکی رفسنجان دوره 19، مهر 1399، 646-645       خوانندگان مقاله خود را قبل از نگارش آن تصور نمایید Imagine the Readers of your article before writing it   محسن رضائیان[1]   M. Rezaeian     هانا (Hanna)، فصل دهم کتاب خود را با عنوان: «تصور نمودن خوانندگان خود» (Envision Your Readers)، مزین نموده است. وی در این فصل از کتاب خود، به این نکته مهم ...

متن کامل

Comparison of parental anxiety before surgery your children

Introduction: Not only do parents become anxious when their child is admitted, but this feeling may continue after the child is discharged from the hospital. The aim of this study was to evaluate and compare the anxiety of parents of children before surgery. Method: This descriptive cross-sectional study was performed in 2019. In this study, 66 parents (parents) of children were randomly enroll...

متن کامل

Having Your Cake and Eating It Too : The Before and After Tax Efficiencies

Recent studies have shown that adding a short extension to a long-only tax exempt equity portfolio leads to a more efficient portfolio. In this work, we show that these extended mandates are even more effective for investors subject to taxes. Not only is the portfolio more efficiently positioned from a pretax perspective, but it offers significant after-tax benefits from increased loss harvesti...

متن کامل

To prune, or not to prune: exploring the efficacy of pruning for model compression

Model pruning seeks to induce sparsity in a deep neural network’s various connection matrices, thereby reducing the number of nonzero-valued parameters in the model. Recent reports (Han et al., 2015a; Narang et al., 2017) prune deep networks at the cost of only a marginal loss in accuracy and achieve a sizable reduction in model size. This hints at the possibility that the baseline models in th...

متن کامل

Understanding Context Before Using It

This paper presents an attempt to point out some problematic issues about the understanding of context. Although frequently used in cognitive sciences or other disciplines, context stays a very ill-defined concept. Our goal is to identify the main components of the context on the basis of the analysis of a corpus of 150 definitions coming mainly from the web in different domains of cognitive sc...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Lecture Notes in Computer Science

سال: 2022

ISSN: ['1611-3349', '0302-9743']

DOI: https://doi.org/10.1007/978-3-031-20083-0_8